Sublinear Time Orthogonal Tensor Decomposition
نویسندگان
چکیده
A recent work (Wang et. al., NIPS 2015) gives the fastest known algorithms for orthogonal tensor decomposition with provable guarantees. Their algorithm is based on computing sketches of the input tensor, which requires reading the entire input. We show in a number of cases one can achieve the same theoretical guarantees in sublinear time, i.e., even without reading most of the input tensor. Instead of using sketches to estimate inner products in tensor decomposition algorithms, we use importance sampling. To achieve sublinear time, we need to know the norms of tensor slices, and we show how to do this in a number of important cases. For symmetric tensors T = ∑k i=1 λiu ⊗p i with λi > 0 for all i, we estimate such norms in sublinear time whenever p is even. For the important case of p = 3 and small values of k, we can also estimate such norms. For asymmetric tensors sublinear time is not possible in general, but we show if the tensor slice norms are just slightly below ‖T ‖F then sublinear time is again possible. One of the main strengths of our work is empirical in a number of cases our algorithm is orders of magnitude faster than existing methods with the same accuracy.
منابع مشابه
Robust polynomial time tensor decomposition
Tensor decomposition has recently become an invaluable algorithmic primitive. It has seen much use in new algorithms with provable guarantees for fundamental statistics and machine learning problems. In these settings, some low-rank k-tensor A ∑r i 1 a ⊗k i which wewould like to decompose into components a1, . . . , ar ∈ n is often not directly accessible. This could happen for many reasons; a...
متن کاملOrthogonal Tensor Decomposition
In symmetric tensor decomposition one expresses a given symmetric tensor T a sum of tensor powers of a number of vectors: T = v⊗d 1 + · · · + v ⊗d k . Orthogonal decomposition is a special type of symmetric tensor decomposition in which in addition the vectors v1, ..., vk are required to be pairwise orthogonal. We study the properties of orthogonally decomposable tensors. In particular, we give...
متن کاملSPALS: Fast Alternating Least Squares via Implicit Leverage Scores Sampling
Tensor CANDECOMP/PARAFAC (CP) decomposition is a powerful but computationally challenging tool in modern data analytics. In this paper, we show ways of sampling intermediate steps of alternating minimization algorithms for computing low rank tensor CP decompositions, leading to the sparse alternating least squares (SPALS) method. Specifically, we sample the Khatri-Rao product, which arises as a...
متن کاملSymmetric Orthogonal Tensor Decomposition is Trivial
We consider the problem of decomposing a real-valued symmetric tensor as the sum of outer products of real-valued, pairwise orthogonal vectors. Such decompositions do not generally exist, but we show that some symmetric tensor decomposition problems can be converted to orthogonal problems following the whitening procedure proposed by Anandkumar et al. (2012). If an orthogonal decomposition of a...
متن کاملA Counterexample to the Possibility of an Extension of the Eckart-Young Low-Rank Approximation Theorem for the Orthogonal Rank Tensor Decomposition
Earlier work has shown that no extension of the Eckart–Young SVD approximation theorem can be made to the strong orthogonal rank tensor decomposition. Here, we present a counterexample to the extension of the Eckart–Young SVD approximation theorem to the orthogonal rank tensor decomposition, answering an open question previously posed by Kolda [SIAM J. Matrix Anal. Appl., 23 (2001), pp. 243–355].
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016